Goto

Collaborating Authors

 orthonormal basis




Large Data Limits of Laplace Learning for Gaussian Measure Data in Infinite Dimensions

Zhong, Zhengang, Korolev, Yury, Thorpe, Matthew

arXiv.org Machine Learning

Laplace learning is a semi-supervised method, a solution for finding missing labels from a partially labeled dataset utilizing the geometry given by the unlabeled data points. The method minimizes a Dirichlet energy defined on a (discrete) graph constructed from the full dataset. In finite dimensions the asymptotics in the large (unlabeled) data limit are well understood with convergence from the graph setting to a continuum Sobolev semi-norm weighted by the Lebesgue density of the data-generating measure. The lack of the Lebesgue measure on infinite-dimensional spaces requires rethinking the analysis if the data aren't finite-dimensional. In this paper we make a first step in this direction by analyzing the setting when the data are generated by a Gaussian measure on a Hilbert space and proving pointwise convergence of the graph Dirichlet energy.


Simultaneously Solving Infinitely Many LQ Mean Field Games In Hilbert Spaces: The Power of Neural Operators

Firoozi, Dena, Kratsios, Anastasis, Yang, Xuwei

arXiv.org Artificial Intelligence

Traditional mean-field game (MFG) solvers operate on an instance-by-instance basis, which becomes infeasible when many related problems must be solved (e.g., for seeking a robust description of the solution under perturbations of the dynamics or utilities, or in settings involving continuum-parameterized agents.). We overcome this by training neural operators (NOs) to learn the rules-to-equilibrium map from the problem data (``rules'': dynamics and cost functionals) of LQ MFGs defined on separable Hilbert spaces to the corresponding equilibrium strategy. Our main result is a statistical guarantee: an NO trained on a small number of randomly sampled rules reliably solves unseen LQ MFG variants, even in infinite-dimensional settings. The number of NO parameters needed remains controlled under appropriate rule sampling during training. Our guarantee follows from three results: (i) local-Lipschitz estimates for the highly nonlinear rules-to-equilibrium map; (ii) a universal approximation theorem using NOs with a prespecified Lipschitz regularity (unlike traditional NO results where the NO's Lipschitz constant can diverge as the approximation error vanishes); and (iii) new sample-complexity bounds for $L$-Lipschitz learners in infinite dimensions, directly applicable as the Lipschitz constants of our approximating NOs are controlled in (ii).




A Proof of Theorem 2

Neural Information Processing Systems

We can see that the error for the first term is mainly due to the sample approximation. We refer to the second term as the Bias . In this section, we establish the bound on the bias. We are now ready to compute the upper bound. The proof will require several lemmas in its construction, which we now present.


Tensor Train Completion from Fiberwise Observations Along a Single Mode

Sofi, Shakir Showkat, De Lathauwer, Lieven

arXiv.org Machine Learning

Tensor completion is an extension of matrix completion aimed at recovering a multiway data tensor by leveraging a given subset of its entries (observations) and the pattern of observation. The low-rank assumption is key in establishing a relationship between the observed and unobserved entries of the tensor. The low-rank tensor completion problem is typically solved using numerical optimization techniques, where the rank information is used either implicitly (in the rank minimization approach) or explicitly (in the error minimization approach). Current theories concerning these techniques often study probabilistic recovery guarantees under conditions such as random uniform observations and incoherence requirements. However, if an observation pattern exhibits some low-rank structure that can be exploited, more efficient algorithms with deterministic recovery guarantees can be designed by leveraging this structure. This work shows how to use only standard linear algebra operations to compute the tensor train decomposition of a specific type of ``fiber-wise" observed tensor, where some of the fibers of a tensor (along a single specific mode) are either fully observed or entirely missing, unlike the usual entry-wise observations. From an application viewpoint, this setting is relevant when it is easier to sample or collect a multiway data tensor along a specific mode (e.g., temporal). The proposed completion method is fast and is guaranteed to work under reasonable deterministic conditions on the observation pattern. Through numerical experiments, we showcase interesting applications and use cases that illustrate the effectiveness of the proposed approach.


Dimension lower bounds for linear approaches to function approximation

Hsu, Daniel

arXiv.org Artificial Intelligence

This short note presents a linear algebraic approach to proving dimension lower bounds for linear methods that solve $L^2$ function approximation problems. The basic argument has appeared in the literature before (e.g., Barron, 1993) for establishing lower bounds on Kolmogorov $n$-widths. The argument is applied to give sample size lower bounds for kernel methods.


Appendix A Further wavelet details

Neural Information Processing Systems

Using this identity, it is easy to check that the highpass filter must have zero-mean, i.e., X Then Eq. 12 and Eq. 13 provides the sufficient and necessary conditions on the highpass filter to build In this section, we show additional results for the experiments with synthetic data in Sec 4.1 . All experiments were run on an A WS instance of p3.16xlarge for a few days. Fig B4 calculates the distance between the learned wavelets and the groundtruth (DB5) wavelet, defined as in Sec 4.1, as the interpretation penalty varies. For a detailed overview of the data, see the original study [ 50 ]. In order to convert the raw fluorescence images to time-series traces, we use tracking code from previous work [ 52 ].